6 research outputs found

    Design of Interactive Feature Space Construction Protocol

    Get PDF
    Machine learning deals with designing systems that learn from data i.e. automatically improve with experience. Systems gain experience by detecting patterns or regularities and using them for making predictions. These predictions are based on the properties that the system learns from the data. Thus when we say a machine learns, it means it has changed in a way that allows it to perform more efficiently than before. Machine learning is emerging as an important technology for solving a number of applications involving natural language processing applications, medical diagnosis, game playing or financial applications. Wide variety of machine learning approaches have been developed and used for a number of applications. We first review the work done in the field of machine learning and analyze various concepts about machine learning that are applicable to the work presented in this thesis. Next we examine active machine learning for pipelining of an important natural language application i.e. information extraction, in which the task of prediction is carried out in different stages and the output of each stage serves as an input to the next stage. A number of machine learning algorithms have been developed for different applications. However no single machine learning algorithm can be used appropriately for all learning problems. It is not possible to create a general learner for all problems because there are varied types of real world datasets that cannot be handled by a single learner. For this purpose an evaluation of the machine learning algorithms is needed. We present an experiment for the evaluation of various state-of-the-art machine learning algorithms using an interactive machine learning tool called WEKA (Waikato Environment for Knowledge Analysis). Evaluation is carried out with the purpose of finding an optimal solution for a real world learning problemcredit approval used in banks. It is a classification problem. Finally, we present an approach of combining various learners with the aim of increasing their efficiency. We present two experiments that evaluate the machine learning algorithms for efficiency and compare their performance with the new combined approach, for the same classification problem. Later we show the effects of feature selection on the efficiency of our combined approach as well as on other machine learning techniques. The aim of this work is to analyze the techniques that increase the efficiency of the learners

    Framework for Human Computer Interaction for Learning Dialogue Strategies using Controlled Natural Language in Information Systems

    Get PDF
    Spoken Language systems are going to have a tremendous impact in all the real world applications, be it healthcare enquiry, public transportation system or airline booking system maintaining the language ethnicity for interaction among users across the globe. These system have the capability of interacting with the user in di erent languages that the system supports. Normally when a person interacts with another person there are many non-verbal clues which guide the dialogue and all the utterances have a contextual relationship, which manage the dialogue as its mixed by the two speakers. Human Computer Interaction has a wide impact on the design of the applications and has become one of the emerging interest area of the researchers. All of us are witness to an explosive electronic revolution where lots of gadgets and gizmo's have surrounded us, advanced not only in power, design, applications but the ease of access or what we call user friendly interfaces are designed that we can easily use and control all the functionality of the devices. Since speech is one of the most intuitive form of interaction that humans use. It provides potential bene ts such as handfree access to machines, ergonomics and greater e ciency of interaction. Yet, speech-based interfaces design has been an expert job for a long time. Lot of research has been done in building real spoken Dialogue Systems which can interact with humans using voice interactions and help in performing various tasks as are done by humans. Last two decades have seen utmost advanced research in the automatic speech recognition, dialogue management, text to speech synthesis and Natural Language Processing for various applications which have shown positive results. This dissertation proposes to apply machine learning (ML) techniques to the problem of optimizing the dialogue management strategy selection in the Spoken Dialogue system prototype design. Although automatic speech recognition and system initiated dialogues where the system expects an answer in the form of `yes' or `no' have already been applied to Spoken Dialogue Systems( SDS), no real attempt to use those techniques in order to design a new system from scratch has been made. In this dissertation, we propose some novel ideas in order to achieve the goal of easing the design of Spoken Dialogue Systems and allow novices to have access to voice technologies. A framework for simulating and evaluating dialogues and learning optimal dialogue strategies in a controlled Natural Language is proposed. The simulation process is based on a probabilistic description of a dialogue and on the stochastic modelling of both arti cial NLP modules composing a SDS and the user. This probabilistic model is based on a set of parameters that can be tuned from the prior knowledge from the discourse or learned from data. The evaluation is part of the simulation process and is based on objective measures provided by each module. Finally, the simulation environment is connected to a learning agent using the supplied evaluation metrics as an objective function in order to generate an optimal behaviour for the SDS

    Modeling of Security Measurement (Metrics) in an Information System

    Get PDF
    Security metrics and measurement is a sub-field of broader information security field. This field is not new but it got very least and sporadic attention as a result of which it is still in its early stages. The measurement and evaluation of security now became a long standing challenge to the research community. Much of the focus remained towards devising and the application of new and updated protection mechanisms. Measurements in general act as a driving force in decision making. As stated by Lord Kelvin “if you cannot measure it then you cannot improve it”. This principle is also applicable to security measurement of information systems. Even if the necessary and required protection mechanisms are in place still the level of security remains unknown, which limits the decision making capabilities to improve the security of a system. With the increasing reliance on these information systems in general and software systems in particular security measurement has become the most pressing requirement in order to promote and develop the security critical systems in the current networked environment. The resultant indicators of security measurement preferably the quantative indicators act as a basis for the decision making to enhance the security of overall system. The information systems are comprised of various components such as people, hardware, data, network and software. With the fast growing reliance on the software systems, the research reported in this thesis aims to provide a framework using mathematical modeling techniques for evaluation of security of the software systems at the architectural and design phase of the system lifecycle and the derived security metrics on a controlled scale from the proposed framework. The proposed security evaluation framework is independent of the programing language and the platform used in developing the system and also is applicable from small desktop application to large complex distributed software. The validation process of security metrics is the most challenging part of the security metrics field. In this thesis we have conducted the exploratory empirical evaluation on a running system to validate the derived security metrics and the measurement results. To make the task easy we have transformed the proposed security evaluation into algorithmic form which increased the applicability of the proposed framework without requiring any expert security knowledge. The motivation of the research is to provide the software development team with a tool to evaluate the level of security of each of the element of the system and the overall system at the early development stages of the system life cycle. In this regard three question “What is to be measured?”, “where (in the system life cycle) to measure?” and “how to measure?” have been answered in the thesis. Since the field of security metrics and measurements is still in the its early stages, the first part of the thesis investigates and analyzes the basic terminologies , taxonomies and major efforts made towards security metrics based on the literature survey. Answering the second question “Where (in the system life cycle) to measure security”, the second part of the thesis analyzes the secure software development processes (SSDPs) followed and identifies the key stages of the system’s life cycle where the evaluation of security is necessary. Answering the question 1 and 2, “What is to be measured “and “How to measure”, third part of the thesis presents a security evaluation framework aimed at the software architecture and design phase using mathematical modeling techniques. In the proposed framework, the component based architecture and design (CBAD) using UML 2.0 component modeling techniques has been adopted. Further in part 3 of the thesis present the empirical evaluation of the proposed framework to validate and analyze the applicability and feasibility of the proposed security metrics. Our effort is to get the focus of the software development community to focus on the security evaluation in the software development process in order to take the early decisions regarding the security of the overall system

    Software Reliability Growth Models from the Perspective of Learning Effects and Change-Point.

    Get PDF
    Increased attention towards reliability of software systems has led to the thorough analysis of the process of reliability growth for prediction and assessment of software reliability in the testing or debugging phase. With many frameworks available in terms of the underlying probability distributions like Poisson process, Non-Homogeneous Poisson Process (NHPP), Weibull, etc, many researchers have developed models using the Non-Homogeneous Poisson Process (NHPP) analytical framework. The behavior of interest, usually, is S-shaped or exponential shaped. S-shaped behavior could relate more closely to the human learning. The need to develop different models stems from the fact that nature of the underlying environment, learning effect acquisition during testing, resource allocations, application and the failure data itself vary. There is no universal model that fits everywhere to be called an Oracle. Learning effects that stem from the experiences of the testing or debugging staff have been considered for the growth of reliability. Learning varies over time and this asserts need for conduct of more research for study of learning effects.Digital copy of ThesisUniversity of Kashmi

    Evaluating Effectiveness of Software Testing Techniques with Emphasis on Enhancing Software Reliability

    Get PDF
    Software testing is one of the most widely known and essential field in software engineering. The purpose of software testing is not only to reveal defects and eliminate them but also to serve as a tool for verification, validation and certification. Defection detection and increasing reliability are the two main goals of software testing. For decades, researchers have been inventing new techniques to test software. However, no testing technique will ever be a solution for all types of software defects. At present, we have very limited information of software testing techniques effectiveness and efficiency. Therefore, while researchers should continue to develop new testing techniques, they also need to deeply understand the abilities and limitations of existing techniques. We need to know what types of defects a particular technique can be expected to find and at what cost. We have to check whether testing technique effectiveness and efficiency depends on program to which it is applied, subject who applies it, the number of faults in the program or the type of faults in the program. However it is not sufficient if testing techniques are only compared on fault detecting ability. They should also be evaluated to check which among them enhances reliability. The research in this thesis aims at evaluating software testing techniques in terms of effectiveness in detecting software defects, and the ability to increase the reliability of the software. The research in this thesis falls within empirical method research on the verification and validation process with a focus on software testing techniques evaluation. The work in this thesis links both research and practice and aims to continue building empirical knowledge in the field of software engineering. The first part of this thesis surveys and analyzes empirical studies on evaluation of testing techniques. Issues with the current evaluation of software testing techniques are identified. Building upon this, we present an evaluation framework (a set of guidelines) for experiments which evaluate the software testing techniques. In addition, we also proposed a uniform classification of software testing techniques and identified a set of factors which helps us in selecting an appropriate testing technique. The second part of the thesis presents an experiment which evaluates and com- pares three defect detection techniques to evaluate the effectiveness of the software testing techniques in terms of defect detection. Moreover, we also evaluated the efficiency of these techniques. The dependence of the effectiveness and efficiency on the various parameters like programs, subjects and defects is also investigated in the experiment. The third and final part of this thesis presents an experiment which evaluates and compares three defect detection techniques for reliability using a novel method. The efficiency of these techniques is also evaluated. Our effort is to provide evidence that will help testing and research community to understand the effectiveness and efficiency of software testing techniques in terms of defect detection and reliability and their dependence on various factors. The ultimate goal of our work is to move software engineering from a craft towards an engineering discipline
    corecore